Recall (machine learning)

A metric for classification models that answers the following question:1

When ground truth was the positive class, what percentage of predictions did the model correctly identify as the positive class?

Formula

Recall is the proportion of all actual positives that were classified correctly as positives:

correctly classified actual positivesall actual positives\frac{\text{correctly classified actual positives}}{\text{all actual positives}}

which means:

Recall=TPTP+FN\text{Recall} = \frac{TP}{TP + FN}

where:

Precision improves as false positives decrease, while recall improves when false negatives decrease. They often show an inverse relationship, where improving one of them worsens the other. (See also F1 score)

When (not) to use

Use when false negative are more expensive than false positives.2

Recall is particularly useful for determining the predictive power of classification models in which the positive class is rare. Recall is a much more useful metric for class-imbalanced datasets than accuracy.1 But the number of actual positives is very, very low, say 1-2 examples in total, recall is less meaningful and less useful as a metric.2

See also

Footnotes

  1. developers.google.com/machine-learning/glossary#recall 2

  2. ML crash course - Classification 2

2024 © ak